spatiotemporal lstm
- North America > United States > California > Los Angeles County > Long Beach (0.04)
- Asia > China > Guangdong Province > Guangzhou (0.04)
Reviews: PredRNN: Recurrent Neural Networks for Predictive Learning using Spatiotemporal LSTMs
Here, the authors focus on the maximum likelihood output, but it would be helpful for comparison with prior work to also report the likelihood. Additionally, the computational complexity is mentioned as an advantage of this model, but no detailed analysis or comparison is performed so its hard to know how this compares computational complexity with prior work. Minor notational suggestion: It might be easier for the reader to follow if you use M instead of C for the cell state in equation 3 so that the connection with equation 4 is clearer.
- North America > United States > California > Los Angeles County > Long Beach (0.04)
- Asia > China > Guangdong Province > Guangzhou (0.04)
PredRNN: Recurrent Neural Networks for Predictive Learning using Spatiotemporal LSTMs
Wang, Yunbo, Long, Mingsheng, Wang, Jianmin, Gao, Zhifeng, Yu, Philip S.
The predictive learning of spatiotemporal sequences aims to generate future images by learning from the historical frames, where spatial appearances and temporal variations are two crucial structures. This architecture is enlightened by the idea that spatiotemporal predictive learning should memorize both spatial appearances and temporal variations in a unified memory pool. Concretely, memory states are no longer constrained inside each LSTM unit. Instead, they are allowed to zigzag in two directions: across stacked RNN layers vertically and through all RNN states horizontally. The core of this network is a new Spatiotemporal LSTM (ST-LSTM) unit that extracts and memorizes spatial and temporal representations simultaneously.
PredRNN: Recurrent Neural Networks for Predictive Learning using Spatiotemporal LSTMs
Wang, Yunbo, Long, Mingsheng, Wang, Jianmin, Gao, Zhifeng, Yu, Philip S.
The predictive learning of spatiotemporal sequences aims to generate future images by learning from the historical frames, where spatial appearances and temporal variations are two crucial structures. This paper models these structures by presenting a predictive recurrent neural network (PredRNN). This architecture is enlightened by the idea that spatiotemporal predictive learning should memorize both spatial appearances and temporal variations in a unified memory pool. Concretely, memory states are no longer constrained inside each LSTM unit. Instead, they are allowed to zigzag in two directions: across stacked RNN layers vertically and through all RNN states horizontally. The core of this network is a new Spatiotemporal LSTM (ST-LSTM) unit that extracts and memorizes spatial and temporal representations simultaneously. PredRNN achieves the state-of-the-art prediction performance on three video prediction datasets and is a more general framework, that can be easily extended to other predictive learning tasks by integrating with other architectures.
- North America > United States > California > Los Angeles County > Long Beach (0.04)
- Asia > China > Guangdong Province > Guangzhou (0.04)